Fast Randomized Algorithms for Convex Optimization and Statistical Estimation
نویسنده
چکیده
Fast Randomized Algorithms for Convex Optimization and Statistical Estimation
منابع مشابه
Randomized Smoothing for Stochastic Optimization
We analyze convergence rates of stochastic optimization algorithms for nonsmooth convex optimization problems. By combining randomized smoothing techniques with accelerated gradient methods, we obtain convergence rates of stochastic optimization procedures, both in expectation and with high probability, that have optimal dependence on the variance of the gradient estimates. To the best of our k...
متن کاملLinear Time Varying MPC Based Path Planning of an Autonomous Vehicle via Convex Optimization
In this paper a new method is introduced for path planning of an autonomous vehicle. In this method, the environment is considered cluttered and with some uncertainty sources. Thus, the state of detected object should be estimated using an optimal filter. To do so, the state distribution is assumed Gaussian. Thus the state vector is estimated by a Kalman filter at each time step. The estimation...
متن کاملConstant Nullspace Strong Convexity and Fast Convergence of Proximal Methods under High-Dimensional Settings
State of the art statistical estimators for high-dimensional problems take the form of regularized, and hence non-smooth, convex programs. A key facet of these statistical estimation problems is that these are typically not strongly convex under a high-dimensional sampling regime when the Hessian matrix becomes rankdeficient. Under vanilla convexity however, proximal optimization methods attain...
متن کاملEfficient algorithms for online convex optimization and their applications
In this thesis we study algorithms for online convex optimization and their relation to approximate optimization. In the first part, we propose a new algorithm for a general online optimization framework called online convex optimization. Whereas previous efficient algorithms are mostly gradient-descent based, the new algorithm is inspired by the Newton-Raphson method for convex optimization, a...
متن کاملFast Stochastic Variance Reduced ADMM for Stochastic Composition Optimization
We consider the stochastic composition optimization problem proposed in [17], which has applications ranging from estimation to statistical and machine learning. We propose the first ADMM-based algorithm named com-SVRADMM, and show that com-SVR-ADMM converges linearly for strongly convex and Lipschitz smooth objectives, and has a convergence rate of O(logS/S), which improves upon the O(S−4/9) r...
متن کامل